14 research outputs found

    Optimal Quantisation of Probability Measures Using Maximum Mean Discrepancy

    Get PDF
    Several researchers have proposed minimisation of maximum mean discrepancy (MMD) as a method to quantise probability measures, i.e., to approximate a distribution by a representative point set. We consider sequential algorithms that greedily minimise MMD over a discrete candidate set. We propose a novel non-myopic algorithm and, in order to both improve statistical efficiency and reduce computational cost, we investigate a variant that applies this technique to a mini-batch of the candidate set at each iteration. When the candidate points are sampled from the target, the consistency of these new algorithms—and their mini-batch variants—is established. We demonstrate the algorithms on a range of important computational problems, including optimisation of nodes in Bayesian cubature and the thinning of Markov chain output

    Sharp Gaussian Approximation Bounds for Linear Systems with α-stable Noise

    Get PDF
    We report the results of several theoretical studies into the convergence rate for certain random series representations of α-stable random variables, which are motivated by and find application in modelling heavy-tailed noise in time series analysis, inference, and stochastic processes. The use of α-stable noise distributions generally leads to analytically intractable inference problems. The particular version of the Poisson series representation invoked here implies that the resulting distributions are “conditionally Gaussian,” for which inference is relatively straightforward, although an infinite series is still involved. Our approach is to approximate the residual (or “tail”) part of the series from some point, c > 0, say, to ∞, as a Gaussian random variable. Empirically, this approximation has been found to be very accurate for large c. We study the rate of convergence, as c → ∞, of this Gaussian approximation. This allows the selection of appropriate truncation parameters, so that a desired level of accuracy for the approximate model can be achieved. Explicit, nonasymptotic bounds are obtained for the Kolmogorov distance between the relevant distribution functions, through the application of probability-theoretic tools. The theoretical results obtained are found to be in very close agreement with numerical results obtained in earlier work

    Optimal Thinning of MCMC Output

    Get PDF
    The use of heuristics to assess the convergence and compress the output of Markov chain Monte Carlo can be sub-optimal in terms of the empirical approximations that are produced. Typically a number of the initial states are attributed to "burn in" and removed, whilst the remainder of the chain is "thinned" if compression is also required. In this paper we consider the problem of retrospectively selecting a subset of states, of fixed cardinality, from the sample path such that the approximation provided by their empirical distribution is close to optimal. A novel method is proposed, based on greedy minimisation of a kernel Stein discrepancy, that is suitable for problems where heavy compression is required. Theoretical results guarantee consistency of the method and its effectiveness is demonstrated in the challenging context of parameter inference for ordinary differential equations. Software is available in the Stein Thinning package in Python, R and MATLAB.Comment: To appear in the Journal of the Royal Statistical Society, Series B, 2021

    The LĂ©vy State Space Model

    No full text

    Nonasymptotic Gaussian Approximation for Inference with Stable Noise

    No full text
    The results of a series of theoretical studies are reported, examining the convergence rate for different approximate representations of α\alpha-stable distributions. Although they play a key role in modelling random processes with jumps and discontinuities, the use of α\alpha-stable distributions in inference often leads to analytically intractable problems. The LePage series, which is a probabilistic representation employed in this work, is used to transform an intractable, infinite-dimensional inference problem into a conditionally Gaussian parametric problem. A major component of our approach is the approximation of the tail of this series by a Gaussian random variable. Standard statistical techniques, such as Expectation-Maximization, Markov chain Monte Carlo, and Particle Filtering, can then be applied. In addition to the asymptotic normality of the tail of this series, we establish explicit, nonasymptotic bounds on the approximation error. Their proofs follow classical Fourier-analytic arguments, using Ess\'{e}en's smoothing lemma. Specifically, we consider the distance between the distributions of: (i)(i)~the tail of the series and an appropriate Gaussian; (ii)(ii)~the full series and the truncated series; and (iii)(iii)~the full series and the truncated series with an added Gaussian term. In all three cases, sharp bounds are established, and the theoretical results are compared with the actual distances (computed numerically) in specific examples of symmetric α\alpha-stable distributions. This analysis facilitates the selection of appropriate truncations in practice and offers theoretical guarantees for the accuracy of resulting estimates. One of the main conclusions obtained is that, for the purposes of inference, the use of a truncated series together with an approximately Gaussian error term has superior statistical properties and is likely a preferable choice in practice.Comment: V1: 41 pages, 16 figures. V2: Text typos fixed; redundant figures from main text and appendices removed; added references in section I; changed section VI and its proofs in Appendices C-D-E; improved section IX; removed section X (Discussion and Conclusion), reference style changed. V3: title in the metadata updated. V4: Updated Theorem 1 and its proof, global revisio

    Sharp Gaussian Approximation Bounds for Linear Systems with α-stable Noise

    No full text
    corecore